首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   774篇
  免费   24篇
  国内免费   1篇
电工技术   4篇
化学工业   146篇
金属工艺   5篇
机械仪表   12篇
建筑科学   22篇
能源动力   25篇
轻工业   59篇
水利工程   8篇
无线电   106篇
一般工业技术   91篇
冶金工业   14篇
原子能技术   5篇
自动化技术   302篇
  2024年   1篇
  2023年   5篇
  2022年   13篇
  2021年   18篇
  2020年   12篇
  2019年   12篇
  2018年   15篇
  2017年   17篇
  2016年   23篇
  2015年   27篇
  2014年   31篇
  2013年   61篇
  2012年   54篇
  2011年   55篇
  2010年   45篇
  2009年   59篇
  2008年   52篇
  2007年   53篇
  2006年   43篇
  2005年   31篇
  2004年   23篇
  2003年   23篇
  2002年   18篇
  2001年   9篇
  2000年   7篇
  1999年   9篇
  1998年   11篇
  1997年   11篇
  1996年   11篇
  1995年   8篇
  1994年   9篇
  1993年   3篇
  1992年   5篇
  1991年   4篇
  1990年   3篇
  1989年   2篇
  1988年   3篇
  1987年   3篇
  1986年   1篇
  1985年   1篇
  1984年   1篇
  1980年   1篇
  1977年   1篇
  1970年   1篇
  1965年   1篇
  1958年   1篇
  1957年   2篇
排序方式: 共有799条查询结果,搜索用时 265 毫秒
31.
The geometric element transformation method (GETMe) is a geometry-based smoothing method for mixed and non-mixed meshes. It is based on a simple geometric transformation applicable to elements bounded by polygons with an arbitrary number of nodes. The transformation, if applied iteratively, leads to a regularization of the polygons. Global mesh smoothing is accomplished by averaging the new node positions obtained by local element transformations. Thereby, the choice of transformation parameters as well as averaging weights can be based on the element quality which leads to high quality results. In this paper, a concept of an enhanced transformation approach is presented and a proof for the regularizing effect of the transformation based on eigenpolygons is given. Numerical examples confirm that the GETMe approach leads to superior mesh quality if compared to other geometry-based methods. In terms of quality it can even compete with optimization-based techniques, despite being conceptually significantly simpler.  相似文献   
32.
SQL injection attacks involve the construction of application input data that will result in the execution of malicious SQL statements. Many web applications are prone to SQL injection attacks. This paper proposes a novel methodology of preventing this kind of attacks by placing a secure database driver between the application and its underlying relational database management system. To detect an attack, the driver uses stripped-down SQL queries and stack traces to create SQL statement signatures that are then used to distinguish between injected and legitimate queries. The driver depends neither on the application nor on the RDBMS and can be easily retrofitted to any system. We have developed a tool, SDriver, that implements our technique and used it on several web applications with positive results.  相似文献   
33.
Dimitris  Nikos  Costas   《Computers & Security》2009,28(7):578-591
Any application or service utilizing the Internet is exposed to both general Internet attacks and other specific ones. Most of the times the latter are exploiting a vulnerability or misconfiguration in the provided service and/or in the utilized protocol itself. Consequently, the employment of critical services, like Voice over IP (VoIP) services, over the Internet is vulnerable to such attacks and, on top of that, they offer a field for new attacks or variations of existing ones. Among the various threats–attacks that a service provider should consider are the flooding attacks, at the signaling level, which are very similar to those against TCP servers but have emerged at the application level of the Internet architecture. This paper examines flooding attacks against VoIP architectures that employ the Session Initiation Protocol (SIP) as their signaling protocol. The focus is on the design and implementation of the appropriate detection method. Specifically, a bloom filter based monitor is presented and a new metric, named session distance, is introduced in order to provide an effective protection scheme against flooding attacks. The proposed scheme is evaluated through experimental test bed architecture under different scenarios. The results of the evaluation demonstrate that the required time to detect such an attack is negligible and also that the number of false alarms is close to zero.  相似文献   
34.
In the Internet era, users’ fundamental privacy and anonymity rights have received significant research and regulatory attention. This is not only a result of the exponential growth of data that users generate when accomplishing their daily task by means of computing devices with advanced capabilities, but also because of inherent data properties that allow them to be linked with a real or soft identity. Service providers exploit these facts for user monitoring and identification, albeit impacting users’ anonymity, based mainly on personal identifiable information or on sensors that generate unique data to provide personalized services. In this paper, we report on the feasibility of user identification using general system features like memory, CPU and network data, as provided by the underlying operating system. We provide a general framework based on supervised machine learning algorithms both for distinguishing users and informing them about their anonymity exposure. We conduct a series of experiments to collect trial datasets for users’ engagement on a shared computing platform. We evaluate various well-known classifiers in terms of their effectiveness in distinguishing users, and we perform a sensitivity analysis of their configuration setup to discover optimal settings under diverse conditions. Furthermore, we examine the bounds of sampling data to eliminate the chances of user identification and thus promote anonymity. Overall results show that under certain configurations users’ anonymity can be preserved, while in other cases users’ identification can be inferred with high accuracy, without relying on personal identifiable information.  相似文献   
35.
With the advent of the information and related emerging technologies, such as RFID, small size sensors and sensor networks or, more generally, product embedded information devices (PEID), a new generation of products called smart or intelligent products is available in the market.Although various definitions of intelligent products have been proposed, we introduce a new definition of the notion of Intelligent Product inspired by what happens in nature with us as human beings and the way we develop intelligence and knowledge. We see an intelligent product as a product system which contains sensing, memory, data processing, reasoning and communication capabilities at four intelligence levels. This future generations of Intelligent Products will need new Product Data Technologies allowing the seamless interoperability of systems and exchange of not only Static but of Dynamic Product Data as well. Actual standards for PDT cover only lowest intelligence of today’s products. In this context, we try to shape the actual state and a possible future of the Product Data Technologies from a Closed-Loop Product Lifecycle Management (C-L PLM) perspective.Our approach is founded in recent findings of the FP6 IP 507100 project PROMISE and follow-up research work. Standards of the STEP family, covering the product lifecycle to a certain extend (PLCS) as well as MIMOSA and ISO 15926 are discussed together with more recent technologies for the management of ID and sensor data such as EPCglobal, OGC-SWE and relevant PROMISE propositions for standards.Finally, the first efforts towards ontology based semantic standards for product lifecycle management and associated knowledge management and sharing are presented and discussed.  相似文献   
36.
An improved version of the function estimation program GDF is presented. The main enhancements of the new version include: multi-output function estimation, capability of defining custom functions in the grammar and selection of the error function. The new version has been evaluated on a series of classification and regression datasets, that are widely used for the evaluation of such methods. It is compared to two known neural networks and outperforms them in 5 (out of 10) datasets.

Program summary

Title of program: GDF v2.0Catalogue identifier: ADXC_v2_0Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADXC_v2_0.htmlProgram obtainable from: CPC Program Library, Queen's University, Belfast, N. IrelandLicensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.htmlNo. of lines in distributed program, including test data, etc.: 98 147No. of bytes in distributed program, including test data, etc.: 2 040 684Distribution format: tar.gzProgramming language: GNU C++Computer: The program is designed to be portable in all systems running the GNU C++ compilerOperating system: Linux, Solaris, FreeBSDRAM: 200000 bytesClassification: 4.9Does the new version supersede the previous version?: YesNature of problem: The technique of function estimation tries to discover from a series of input data a functional form that best describes them. This can be performed with the use of parametric models, whose parameters can adapt according to the input data.Solution method: Functional forms are being created by genetic programming which are approximations for the symbolic regression problem.Reasons for new version: The GDF package was extended in order to be more flexible and user customizable than the old package. The user can extend the package by defining his own error functions and he can extend the grammar of the package by adding new functions to the function repertoire. Also, the new version can perform function estimation of multi-output functions and it can be used for classification problems.Summary of revisions: The following features have been added to the package GDF:
Multi-output function approximation. The package can now approximate any function . This feature gives also to the package the capability of performing classification and not only regression.
User defined function can be added to the repertoire of the grammar, extending the regression capabilities of the package. This feature is limited to 3 functions, but easily this number can be increased.
Capability of selecting the error function. The package offers now to the user apart from the mean square error other error functions such as: mean absolute square error, maximum square error. Also, user defined error functions can be added to the set of error functions.
More verbose output. The main program displays more information to the user as well as the default values for the parameters. Also, the package gives to the user the capability to define an output file, where the output of the gdf program for the testing set will be stored after the termination of the process.
Additional comments: A technical report describing the revisions, experiments and test runs is packaged with the source code.Running time: Depending on the train data.  相似文献   
37.
We present a new methodology for agent modeling that is scalable and efficient. It is based on the integration of nonlinear dynamical systems and kinetic data structures. The method consists of three layers, which together model 3D agent steering, crowds and flocks among moving and static obstacles. The first layer, the local layer employs nonlinear dynamical systems theory to models low-level behaviors. It is fast and efficient, and it does not depend on the total number of agents in the environment. This dynamical systems-based approach also allows us to establish continuous numerical parameters for modifying each agent's behavior. The second layer, a global environment layer consists of a specifically designed kinetic data structure to track efficiently the immediate environment of each agent and know which obstacles/agents are near or visible to the given agent. This layer reduces the complexity in the local layer. In the third layer, a global planning layer, the problem of target tracking is generalized in a way that allows navigation in maze-like terrains, avoidance of local minima and cooperation between agents. We implement this layer based on two approaches that are suitable for different applications: One approach is to track the closest single moving or static target; the second is to use a pre-specified vector field, which may be generated automatically (with harmonic functions, for example) or based on user input to achieve the desired output. We also discuss how hybrid systems concepts for global planning can capitalize on both our layered approach and the continuous, reactive nature of our agent steering.

We demonstrate the power of the approach through a series of experiments simulating single/multiple agents and crowds moving towards moving/static targets in complex environments.  相似文献   

38.
We propose a forward sequential feature selection scheme based on k‐means clustering algorithm to derive the feature subset that classifies best the time series data base, according to the criterion of the corrected Rand index. Moreover, we investigate the effect of the standardization scheme on the feature selection and propose a standardization given by the transform to standard Gaussian distribution. Our interest in this work is in classification of oscillating dynamical systems on the basis of measures computed on time series from these systems. The features to be selected are measures of linear and non‐linear analysis of time series, such as auto‐correlation and Lyapunov exponents, as well as oscillation characteristics, such as the mean magnitude of peaks. Simulations on known oscillating deterministic and stochastic systems showed that, for repeated realizations of the same classification task, the proposed feature selection scheme selected very often the same best feature subset, giving high classification accuracy for any standardization. We found that, regardless of the standardization, the highest classification accuracy could be obtained with a small feature subset, containing most frequently an oscillating‐related feature. The same setting was applied to records of epileptic electroencephalogram signals, giving varying results and dependent on the standardization.  相似文献   
39.
40.
In this paper we study the ramification problem in the setting of temporal databases. Standard solutions from the literature on reasoning about action are inadequate because they rely on the assumption that fluents persist, and because actions have effects on the next situation only. In this paper we provide a solution to the ramification problem based on an extension of the situation calculus and the work of McCain and Turner. More specifically, we study the case where there are conflicting effects of an action, a particularly complex problem. Also we present a tool which implements the proposed solution.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号